Recovering Structured Probability Matrices
نویسندگان
چکیده
We consider the problem of accurately recovering a matrix B of size M ×M , which represents a probability distribution over M2 outcomes, given access to an observed matrix of “counts” generated by taking independent samples from the distribution B. How can structural properties of the underlying matrix B be leveraged to yield computationally efficient and information theoretically optimal reconstruction algorithms? When can accurate reconstruction be accomplished in the sparse data regime? This basic problem lies at the core of a number of questions that are currently being considered by different communities, including building recommendation systems and collaborative filtering in the sparse data regime, community detection in sparse random graphs, learning structured models such as topic models or hidden Markov models, and the efforts from the natural language processing community to compute “word embeddings”. Many aspects of this problem—both in terms of learning and property testing/estimation and on both the algorithmic and information theoretic sides—remain open. Our results apply to the setting where B has a low rank structure. For this setting, we propose an efficient (and practically viable) algorithm that accurately recovers the underlying M × M matrix using Θ(M) samples (where we assume the rank is a constant). This linear sample complexity is optimal, up to constant factors, in an extremely strong sense: even testing basic properties of the underlying matrix (such as whether it has rank 1 or 2) requires Ω(M) samples. Additionally, we provide an even stronger lower bound showing that distinguishing whether a sequence of observations were drawn from the uniform distribution overM observations versus being generated by a well-conditioned Hidden Markov Model with two hidden states requires Ω(M) observations, while our positive results for recovering B immediately imply that Ω(M) observations suffice to learn such an HMM. This lower bound precludes sublinear-sample hypothesis tests for basic properties, such as identity or uniformity, as well as sublinear sample estimators for quantities such as the entropy rate of HMMs. 1998 ACM Subject Classification I.2.6 Learning
منابع مشابه
Noise Thresholds for Spectral Clustering
Although spectral clustering has enjoyed considerable empirical success in machine learning, its theoretical properties are not yet fully developed. We analyze the performance of a spectral algorithm for hierarchical clustering and show that on a class of hierarchically structured similarity matrices, this algorithm can tolerate noise that grows with the number of data points while still perfec...
متن کاملSymmetric Toeplitz-Structured Compressed Sensing Matrices
Abstract How to construct a suitable measurement matrix is still an open question in compressed sensing. A significant part of the recent work is that the measurement matrices are not completely random on the entries but exhibit considerable structure. In this paper, we proved that the symmetric Toeplitz matrix and its transforms can be used as measurement matrix and recovery signal with high p...
متن کاملStructured Random Matrices
Random matrix theory is a well-developed area of probability theory that has numerous connections with other areas of mathematics and its applications. Much of the literature in this area is concerned with matrices that possess many exact or approximate symmetries, such as matrices with i.i.d. entries, for which precise analytic results and limit theorems are available. Much less well understoo...
متن کاملA Two-pronged Progress in Structured Dense Matrix Vector Multiplication
Matrix-vector multiplication is one of the most fundamental computing primitives. Given a matrix A ∈ FN×N and a vector b ∈ F , it is known that in the worst case Θ(N) operations over F are needed to compute Ab. Many types of structured matrices do admit faster multiplication. However, even given a matrix A that is known to have this property, it is hard in general to recover a representation of...
متن کاملThe probability of primeness for specially structured polynomial matrices over finite fields with applications to linear systems and convolutional codes
We calculate the probability that random polynomial matrices over a finite field with certain structures are right prime or left prime, respectively. In particular, we give an asymptotic formula for the probability that finitely many nonsingular polynomial matrices are mutually left coprime. These results are used to estimate the number of reachable and observable linear systems as well as the ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2018